# Efficient CPU Inference
Bge Large Zh V1.5 Gguf
MIT
BAAI/bge-large-zh-v1.5 embedding models in GGUF format, both quantized and non-quantized, optimized for llama.cpp, delivering significant speedup on CPU with minimal precision loss.
Text Embedding
B
CompendiumLabs
1,213
12
Bge Base En V1.5 Gguf
MIT
This project provides the BGE embedding model stored in GGUF format, which is suitable for use with llama.cpp and offers better performance than transformers.
Text Embedding
B
CompendiumLabs
1,108
5
Featured Recommended AI Models